87 research outputs found

    Echoes of the spoken past: how auditory cortex hears context during speech perception.

    Get PDF
    What do we hear when someone speaks and what does auditory cortex (AC) do with that sound? Given how meaningful speech is, it might be hypothesized that AC is most active when other people talk so that their productions get decoded. Here, neuroimaging meta-analyses show the opposite: AC is least active and sometimes deactivated when participants listened to meaningful speech compared to less meaningful sounds. Results are explained by an active hypothesis-and-test mechanism where speech production (SP) regions are neurally re-used to predict auditory objects associated with available context. By this model, more AC activity for less meaningful sounds occurs because predictions are less successful from context, requiring further hypotheses be tested. This also explains the large overlap of AC co-activity for less meaningful sounds with meta-analyses of SP. An experiment showed a similar pattern of results for non-verbal context. Specifically, words produced less activity in AC and SP regions when preceded by co-speech gestures that visually described those words compared to those words without gestures. Results collectively suggest that what we 'hear' during real-world speech perception may come more from the brain than our ears and that the function of AC is to confirm or deny internal predictions about the identity of sounds

    Speech Perception under the Tent: A Domain-general Predictive Role for the Cerebellum

    Get PDF
    The role of the cerebellum in speech perception remains a mystery. Given its uniform architecture, we tested the hypothesis that it implements a domain-general predictive mechanism whose role in speech is determined by connectivity. We collated all neuroimaging studies reporting cerebellar activity in the Neurosynth database (n = 8206). From this set, we found all studies involving passive speech and sound perception (n = 72, 64% speech, 12.5% sounds, 12.5% music, and 11% tones) and speech production and articulation (n = 175). Standard and coactivation neuroimaging meta-analyses were used to compare cerebellar and associated cortical activations between passive perception and production. We found distinct regions of perception- and production-related activity in the cerebellum and regions of perception–production overlap. Each of these regions had distinct patterns of cortico-cerebellar connectivity. To test for domain-generality versus specificity, we identified all psychological and task-related terms in the Neurosynth database that predicted activity in cerebellar regions associated with passive perception and production. Regions in the cerebellum activated by speech perception were associated with domain-general terms related to prediction. One hallmark of predictive processing is metabolic savings (i.e., decreases in neural activity when events are predicted). To test the hypothesis that the cerebellum plays a predictive role in speech perception, we examined cortical activation between studies reporting cerebellar activation and those without cerebellar activation during speech perception. When the cerebellum was active during speech perception, there was far less cortical activation than when it was inactive. The results suggest that the cerebellum implements a domain-general mechanism related to prediction during speech perception

    Anxiety and amygdala connectivity during movie-watching

    Get PDF
    Rodent and human studies have implicated an amygdala-prefrontal circuit during threat processing. One possibility is that while amygdala activity underlies core features of anxiety (e.g. detection of salient information), prefrontal cortices (i.e. dorsomedial prefrontal/anterior cingulate cortex) entrain its responsiveness. To date, this has been established in tightly controlled paradigms (predominantly using static face perception tasks) but has not been extended to more naturalistic settings. Consequently, using ‘movie fMRI’—in which participants watch ecologically-rich movie stimuli rather than constrained cognitive tasks—we sought to test whether individual differences in anxiety correlate with the degree of face-dependent amygdala-prefrontal coupling in two independent samples. Analyses suggested increased face-dependent superior parietal activation and decreased speech-dependent auditory cortex activation as a function of anxiety. However, we failed to find evidence for anxiety-dependent connectivity, neither in our stimulus-dependent or -independent analyses. Our findings suggest that work using experimentally constrained tasks may not replicate in more ecologically valid settings and, moreover, highlight the importance of testing the generalizability of neuroimaging findings outside of the original context

    Finding core-periphery structures with node influences

    Get PDF
    Detecting core-periphery structures is one of the outstanding issues in complex network analysis. Various algorithms can identify core nodes and periphery nodes. Recent advances found that many networks from real-world data can be better modeled with multiple pairs of core-periphery nodes. In this study, we propose to use an influence propagation process to detect multiple pairs of core-periphery nodes. In this framework, we assume each node can emit a certain amount of influence and propagate it through the network. Then we identify nodes with large influences as core nodes, and we utilize a maximum influence chain to construct a node-pairing network to determine core-periphery pairs. This approach can take node interactions into consideration and can reduce noises in finding pairs. Experiments on randomly generated networks and real-world networks confirm the efficiency and accuracy of our algorithm

    More than words: word predictability, prosody, gesture and mouth movements in natural language comprehension

    Get PDF
    The ecology of human language is face-to-face interaction, comprising cues such as prosody, co-speech gestures and mouth movements. Yet, the multimodal context is usually stripped away in experiments as dominant paradigms focus on linguistic processing only. In two studies we presented video-clips of an actress producing naturalistic passages to participants while recording their electroencephalogram. We quantified multimodal cues (prosody, gestures, mouth movements) and measured their effect on a well-established electroencephalographic marker of processing load in comprehension (N400). We found that brain responses to words were affected by informativeness of co-occurring multimodal cues, indicating that comprehension relies on linguistic and non-linguistic cues. Moreover, they were affected by interactions between the multimodal cues, indicating that the impact of each cue dynamically changes based on the informativeness of other cues. Thus, results show that multimodal cues are integral to comprehension, hence, our theories must move beyond the limited focus on speech and linguistic processing

    Separating Stimulus-Induced and Background Components of Dynamic Functional Connectivity in Naturalistic fMRI

    Get PDF
    We consider the challenges in extracting stimulus-related neural dynamics from other intrinsic processes and noise in naturalistic functional magnetic resonance imaging (fMRI). Most studies rely on inter-subject correlations (ISC) of low-level regional activity and neglect varying responses in individuals. We propose a novel, data-driven approach based on low-rank plus sparse (L+S) decomposition to isolate stimulus-driven dynamic changes in brain functional connectivity (FC) from the background noise, by exploiting shared network structure among subjects receiving the same naturalistic stimuli. The time-resolved multi-subject FC matrices are modeled as a sum of a low-rank component of correlated FC patterns across subjects, and a sparse component of subject-specific, idiosyncratic background activities. To recover the shared low-rank subspace, we introduce a fused version of principal component pursuit (PCP) by adding a fusion-type penalty on the differences between the columns of the low-rank matrix. The method improves the detection of stimulus-induced group-level homogeneity in the FC profile while capturing inter-subject variability. We develop an efficient algorithm via a linearized alternating direction method of multipliers to solve the fused-PCP. Simulations show accurate recovery by the fused-PCP even when a large fraction of FC edges are severely corrupted. When applied to natural fMRI data, our method reveals FC changes that were time-locked to auditory processing during movie watching, with dynamic engagement of sensorimotor systems for speech-in-noise. It also provides a better mapping to auditory content in the movie than ISC

    Reorganization of the Neurobiology of Language After Sentence Overlearning

    Get PDF
    It is assumed that there are a static set of “language regions” in the brain. Yet, language comprehension engages regions well beyond these, and patients regularly produce familiar “formulaic” expressions when language regions are severely damaged. These suggest that the neurobiology of language is not fixed but varies with experiences, like the extent of word sequence learning. We hypothesized that perceiving overlearned sentences is supported by speech production and not putative language regions. Participants underwent 2 sessions of behavioral testing and functional magnetic resonance imaging (fMRI). During the intervening 15 days, they repeated 2 sentences 30 times each, twice a day. In both fMRI sessions, they “passively” listened to those sentences, novel sentences, and produced sentences. Behaviorally, evidence for overlearning included a 2.1-s decrease in reaction times to predict the final word in overlearned sentences. This corresponded to the recruitment of sensorimotor regions involved in sentence production, inactivation of temporal and inferior frontal regions involved in novel sentence listening, and a 45% change in global network organization. Thus, there was a profound whole-brain reorganization following sentence overlearning, out of “language” and into sensorimotor regions. The latter are generally preserved in aphasia and Alzheimer’s disease, perhaps explaining residual abilities with formulaic expressions in both

    Cefotetan versus Conventional Triple Antibiotic Prophylaxis in Elective Colorectal Cancer Surgery

    Get PDF
    This study examined infectious outcomes in elective colorectal cancer surgery between cefotetan alone or conventional triple antibiotics. From January to December 2007, 461 consecutive primary colorectal cancer patients underwent elective surgery. Group A contained 225 patients who received conventional triple antibiotics (cephalosporin, aminoglycoside and metronidazole) for prophylaxis, and group B contained 236 patients who received cefotetan alone for prophylaxis. Treatment failure was defined as the presence of postoperative infection including surgical-site infection (SSI), anastomotic leakage, and pneumonia or urinary tract infection. The two groups were similar in terms of demographics, American Society of Anesthesiologists (ASA) score, tumour location, stage, surgical approach (conventional open vs. laparoscopy-assisted), and type of operation. The treatment failure rates were 3.1% in Group A and 3.4% in Group B (absolute difference, -0.3%; 95% confidence interval [CI], 0.39 to 3.07, P=0.866), with SSI being the most common reason for failure in both groups (2.7% in Group A and 3.0% in Group B [absolute difference, -0.3%; 95% CI, 0.37 to 3.37, P=0.846]). Cefotetan alone is as effective as triple antibiotics for prophylaxis in primary colorectal cancer patients undergoing elective surgery

    High Frequency of Skin-homing Melanocyte-specific Cytotoxic T Lymphocytes in Autoimmune Vitiligo

    Get PDF
    Vitiligo is an autoimmune condition characterized by loss of epidermal melanocytes. Using tetrameric complexes of human histocompatibility leukocyte antigen (HLA) class I to identify antigen-specific T cells ex vivo, we observed high frequencies of circulating MelanA-specific, A*0201-restricted cytotoxic T lymphocytes (A2–MelanA tetramer+ CTLs) in seven of nine HLA-A*0201–positive individuals with vitiligo. Isolated A2–MelanA tetramer+ CTLs were able to lyse A*0201-matched melanoma cells in vitro and their frequency ex vivo correlated with extent of disease. In contrast, no A2–MelanA tetramer+ CTL could be identified ex vivo in all four A*0201-negative vitiligo patients or five of six A*0201-positive asymptomatic controls. Finally, we observed that the A2–MelanA tetramer+ CTLs isolated from vitiligo patients expressed high levels of the skin homing receptor, cutaneous lymphocyte-associated antigen, which was absent from the CTLs seen in the single A*0201-positive normal control. These data are consistent with a role of skin-homing autoreactive melanocyte-specific CTLs in causing the destruction of melanocytes seen in autoimmune vitiligo. Lack of homing receptors on the surface of autoreactive CTLs could be a mechanism to control peripheral tolerance in vivo

    Cross-Modal Prediction in Speech Perception

    Get PDF
    Speech perception often benefits from vision of the speaker's lip movements when they are available. One potential mechanism underlying this reported gain in perception arising from audio-visual integration is on-line prediction. In this study we address whether the preceding speech context in a single modality can improve audiovisual processing and whether this improvement is based on on-line information-transfer across sensory modalities. In the experiments presented here, during each trial, a speech fragment (context) presented in a single sensory modality (voice or lips) was immediately continued by an audiovisual target fragment. Participants made speeded judgments about whether voice and lips were in agreement in the target fragment. The leading single sensory context and the subsequent audiovisual target fragment could be continuous in either one modality only, both (context in one modality continues into both modalities in the target fragment) or neither modalities (i.e., discontinuous). The results showed quicker audiovisual matching responses when context was continuous with the target within either the visual or auditory channel (Experiment 1). Critically, prior visual context also provided an advantage when it was cross-modally continuous (with the auditory channel in the target), but auditory to visual cross-modal continuity resulted in no advantage (Experiment 2). This suggests that visual speech information can provide an on-line benefit for processing the upcoming auditory input through the use of predictive mechanisms. We hypothesize that this benefit is expressed at an early level of speech analysis
    corecore